我们解决了预测动态场景视频的显着图的问题。我们注意到,从固定数量的观察者的凝视数据重建的地图的准确性随帧而变化,因为它取决于场景的内容。当有有限数量的观察者可用时,此问题尤其如此紧迫。在这种情况下,随着传统的深度学习方法,直接最大限度地减少预测和测量的显着性图之间的差异,导致对嘈杂数据过度接受。我们提出了一种噪声感知培训(NAT)范式,这些培训量量化和占帧特异性凝视数据不准确的不确定性。我们表明NAT在有限的培训数据可用时特别有利,在不同模型,丢失函数和数据集中有实验。我们还引入了基于视频游戏的显着数据集,具有富有的时间语义,每帧多个凝视吸引子。数据集和源代码可在https://github.com/nvlabs/nattacy上获得。
translated by 谷歌翻译
Structural pruning of neural network parameters reduces computation, energy, and memory transfer costs during inference. We propose a novel method that estimates the contribution of a neuron (filter) to the final loss and iteratively removes those with smaller scores. We describe two variations of our method using the first and secondorder Taylor expansions to approximate a filter's contribution. Both methods scale consistently across any network layer without requiring per-layer sensitivity analysis and can be applied to any kind of layer, including skip connections. For modern networks trained on ImageNet, we measured experimentally a high (>93%) correlation between the contribution computed by our methods and a reliable estimate of the true importance. Pruning with the proposed methods leads to an improvement over state-ofthe-art in terms of accuracy, FLOPs, and parameter reduction. On ResNet-101, we achieve a 40% FLOPS reduction by removing 30% of the parameters, with a loss of 0.02% in the top-1 accuracy on ImageNet. Code is available at https://github.com/NVlabs/Taylor_pruning.
translated by 谷歌翻译
以离散特征为特征的现实世界数据集无处不在:从分类调查到临床问卷,从未加权网络到DNA序列。然而,最常见的无监督尺寸还原方法是为连续空间设计的,它们用于离散空间可能会导致错误和偏见。在这封信中,我们介绍了一种算法来推断离散空间中嵌入数据集的固有维度(ID)。我们证明了它在基准数据集上的准确性,并将其应用于分析物种指纹识别的宏基因组数据集,发现了一个令人惊讶的小ID,这表明尽管序列具有高度的序列性,但进化的压力在低维歧管上行动。' 空间。
translated by 谷歌翻译
dadapy是用于分析和表征高维数据歧管的Python软件包。它提供了估计固有维度和概率密度的方法,用于执行基于密度的聚类和比较不同的距离指标。我们回顾包装的主要功能,并在玩具案例和现实世界中的使用中举例说明其使用情况。dadapy可在开源Apache 2.0许可下自由使用。
translated by 谷歌翻译